Learning Hidden Variables in Probabilistic Graphical Models
نویسندگان
چکیده
In the past decades, a great deal of research has focused on learning probabilistic graphical models from data. A serious problem in learning such models is the presence of hidden, or latent, variables. These variables are not observed, yet their interaction with the observed variables has important consequences in terms of representation, inference and prediction. Consequently, numerous works have been directed towards learning probabilistic graphical models with hidden variables. A significantly harder challenge is that of detecting new hidden variables and incorporating them into the network structure. Surprisingly, and despite the recognized importance of hidden variables both in social sciences and the learning community, this problem has received little attention. In this dissertation we explore the problem of learning new hidden variable in real-life domains. We present methods for coping with the different elements that this task encompasses: the detection of new hidden variables; determining the cardinality of new hidden variables; incorporating new hidden variables into learning model. In addition we also address the problem of local maxima that is common in many learning scenarios, and is particularly acute in the presence of hidden variables. We present simple and easy to implement methods that work when training data is relatively plentiful as well as a more elaborate framework that is suitable when the model is particularly complex and the data is sparse. We also consider methods specifically tailored at networks with continuous variables and the added challenges in this scenario. We evaluate all of our methods on both synthetic and real-life data. For the more elaborate methods, we put a particular emphasis on learning complex models with many hidden variables. We demonstrate significant improvement in quantitative prediction on unseen test samples when learning with hidden variables, reaffirming their importance in practice. We also demonstrate that models learned with our methods have hidden variables that are qualitatively appealing and shed light on the learned domain.
منابع مشابه
A Guide to the Literature on Learning Probabilistic Networks from Data
|This literature review discusses di erent methods under the general rubric of learning Bayesian networks from data, and includes some overlapping work on more general probabilistic networks. Connections are drawn between the statistical, neural network, and uncertainty communities, and between the di erentmethodological communities, such as Bayesian, description length, and classical statistic...
متن کاملDiscrete restricted Boltzmann machines
We describe discrete restricted Boltzmann machines: probabilistic graphical models with bipartite interactions between visible and hidden discrete variables. Examples are binary restricted Boltzmann machines and discrete näıve Bayes models. We detail the inference functions and distributed representations arising in these models in terms of configurations of projected products of simplices and ...
متن کاملBuntine : a Guide to the Literature on Learning Graphical Models from Data
| This literature review discusses diierent methods under the general rubric of learning Bayesian networks from data, and includes some overlapping work on more general probabilistic networks. Connections are drawn between the statistical, neural network, and uncertainty communities , and between the diierent methodological communities, such as Bayesian, description length, and classical statis...
متن کاملProbabilistic Language Modeling with Hidden Stochastic Automata
In this paper, we introduce a novel dynamical Bayesian network model for probabilistic language modeling. We refer to this as the Hidden Stochastic Automaton. This model, while based on a generalization of the Hidden Markov model, has qualitatively greater generative power than either the Hidden Markov model itself or any of its existing variants and generalizations. This allows the Hidden Stoc...
متن کاملBuild, Compute, Critique, Repeat: Data Analysis with Latent Variable Models
We survey how to use latent variable models to solve data analysis problems. A latent variable model is a probabilistic model of hidden and observed variables, where the hidden variables encode hidden patterns in our data. We uncover these patterns through the posterior, the conditional distribution of the hidden variables given the observations, which we use to explore, summarize, and form pre...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2004